Dynamic Joint Action Perception for Q-Learning Agents

نویسندگان

  • Nancy Fulda
  • Dan Ventura
چکیده

Q-Iearning is a reinforcement learning alg()rithm that learns expected utilities for stateaction transitions through successive interactions with the environment The algorithm '5 simplicity as well as its convergence properties have made it a popular algorithm for study However; its non-parametric representation of utilities limits its effectiveness in environments with large amounts of perceptual input For example; in multiagent systems; each agent may need to consider the action selections of its counterparts in order to learn effective behaviors This creates a joint action space which grows exponentially with the number of agents in the system In such situations; the Q-learning algorithm quickly becomes intractable This paper presents a new algorithm; Dynamic ,Joint Action Perception; which addresses this problem by allowing each agent to dynamically perceive only those joint action distinctions which are relevant to its own payoffs The result is a smaller joint action space and improved scalability of Q-learning to systems with many agents

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Hierarchical Functional Concepts for Knowledge Transfer among Reinforcement Learning Agents

This article introduces the notions of functional space and concept as a way of knowledge representation and abstraction for Reinforcement Learning agents. These definitions are used as a tool of knowledge transfer among agents. The agents are assumed to be heterogeneous; they have different state spaces but share a same dynamic, reward and action space. In other words, the agents are assumed t...

متن کامل

A Q-learning Based Continuous Tuning of Fuzzy Wall Tracking

A simple easy to implement algorithm is proposed to address wall tracking task of an autonomous robot. The robot should navigate in unknown environments, find the nearest wall, and track it solely based on locally sensed data. The proposed method benefits from coupling fuzzy logic and Q-learning to meet requirements of autonomous navigations. Fuzzy if-then rules provide a reliable decision maki...

متن کامل

User-based Vehicle Route Guidance in Urban Networks Based on Intelligent Multi Agents Systems and the ANT-Q Algorithm

Guiding vehicles to their destination under dynamic traffic conditions is an important topic in the field of Intelligent Transportation Systems (ITS). Nowadays, many complex systems can be controlled by using multi agent systems. Adaptation with the current condition is an important feature of the agents. In this research, formulation of dynamic guidance for vehicles has been investigated based...

متن کامل

Baselines for Joint-Action Reinforcement Learning of Coordination in Cooperative Multi-agent Systems

We report on an investigation of reinforcement learning techniques for the learning of coordination in cooperative multiagent systems. Specifically, we focus on a novel action selection strategy for Q-learning (Watkins 1989). The new technique is applicable to scenarios where mutual observation of actions is not possible. To date, reinforcement learning approaches for such independent agents di...

متن کامل

Informed Initial Policies for Learning in Dec-POMDPs

Decentralized partially observable Markov decision processes (Dec-POMDPs) offer a formal model for planning in cooperative multi-agent systems where agents operate with noisy sensors and actuators and local information. While many techniques have been developed for solving DecPOMDPs exactly and approximately, they have been primarily centralized and reliant on knowledge of the model parameters....

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2003